98 research outputs found

    Learning Utility Surfaces for Movement Selection

    Get PDF
    Humanoid robots are highly redundant systems with respect to the tasks they are asked to perform. This redundancy manifests itself in the number of degrees of freedom of the robot exceeding the dimensionality of the task. Traditionally this redundancy has been utilised through optimal control in the null-space. Some cost function is defined that encodes secondary movement goals and movements are optimised with respect to this functio

    Behaviour Generation in Humanoids by Learning Potential-based Policies from Constrained Motion

    Get PDF
    Movement generation that is consistent with observed or demonstrated behaviour is an efficient way to seed movement planning in complex, high-dimensionalmovement systems like humanoid robots.We present a method for learning potentialbased policies from constrained motion data. In contrast to previous approaches to direct policy learning, our method can combine observations from a variety of contexts where different constraints are in force, to learn the underlying unconstrained policy in form of its potential function. This allows us to generalise and predict behaviour where novel constraints apply. We demonstrate our approach on systems of varying complexity, including kinematic data from the ASIMO humanoid robot with 22 degrees of freedom

    Learning Potential-based Policies from Constrained Motion

    Get PDF
    We present a method for learning potential-based policies from constrained motion data. In contrast to previous approaches to direct policy learning, our method can combine observations from a variety of contexts where different constraints are in force, to learn the underlying unconstrained policy in form of its potential function. This allows us to generalise and predict behaviour where novel constraints apply. As a key ingredient, we first create multiple simple local models of the potential, and align those using an efficient algorithm.We can then detect and discard unsuitable subsets of the data and learn a global model from a cleanly pre-processed training set. We demonstrate our approach on systems of varying complexity, including kinematic data from the ASIMO humanoid robot with 22 degrees of freedom

    Methods for Learning Control Policies from Variable Constraint Demonstrations

    Get PDF
    Many everyday human skills can be framed in terms of performing some task subject to constraints imposed by the task or the environment. Constraints are usually not observable and frequently change between contexts. In this chapter, we explore the problem of learning control policies from data containing variable, dynamic and non-linear constraints on motion. We discuss how an effective approach for doing this is to learn the unconstrained policy in a way that is consistent with the constraints. We then go on to discuss several recent algorithms for extracting policies from movement data, where observations are recorded under variable, unknown constraints. We review a number of experiments testing the performance of these algorithms and demonstrating how the resultant policy models generalise over constraints allowing prediction of behaviour under unseen settings where new constraints apply

    A Novel Method for Learning Policies from Constrained Motion

    Get PDF
    Many everyday human skills can be framed in terms of performing some task subject to constraints imposed by the environment. Constraints are usually unobservable and frequently change between contexts. In this paper, we present a novel approach for learning (unconstrained) control policies from movement data, where observations come from movements under different constraints. As a key ingredient, we introduce a small but highly effective modification to the standard risk functional, allowing us to make a meaningful comparison between the estimated policy and constrained observations. We demonstrate our approach on systems of varying complexity, including kinematic data from the ASIMO humanoid robot with 27 degrees of freedom

    Robust Constraint-consistent Learning

    Get PDF
    Many everyday human skills can be framed in terms of performing some task subject to constraints imposed by the environment. Constraints are usually unobservable and frequently change between contexts. In this paper, we present a novel approach for learning (unconstrained) control policies from movement data, where observations are recorded under different constraint settings. Our approach seamlessly integrates unconstrained and constrained observations by performing hybrid optimisation of two risk functionals. The first is a novel risk functional that makes a meaningful comparison between the estimated policy and constrained observations. The second is the standard risk, used to reduce the expected error under impoverished sets of constraints. We demonstrate our approach on systems of varying complexity, and illustrate its utility for transfer learning of a car washing task from human motion capture data

    A Novel Method for Learning Policies from Variable Constraint Data

    Get PDF
    Many everyday human skills can be framed in terms of performing some task subject to constraints imposed by the environment. Constraints are usually unobservable and frequently change between contexts. In this paper, we present a novel approach for learning (unconstrained) control policies from movement data, where observations come from movements under different constraints. As a key ingredient, we introduce a small but highly effective modification to the standard risk functional, allowing us to make a meaningful comparison between the estimated policy and constrained observations. We demonstrate our approach on systems of varying complexity, including kinematic data from the ASIMO humanoid robot with 27 degrees of freedom, and present results for learning from human demonstration

    Considerations of the Gain Spectrum

    No full text
    We propose the gain spectrum as an efficient means to understand the learning dynamics of MLP gradient based learning and to roughly estimate an upper bound of the necessary network complexity. The feasibility of the approach will be shown by a number of experiments. 1 Introduction The investigation of the learning process of artificial neural networks is an area that still receives a great deal of attention in the research community. Due to the nonlinearity of the governing equations and finite size effects there is still no such thing as a fully automated efficient learning algorithm, if there can be any at all. Recent theoretical results provide guide lines for the construction of learning algorithms and permit some insight into the nature of the adaptation mechanism, but even apparently simple on-line algorithms are far from being sufficiently well understood. Our goal is to contribute to the explanation of the learning process. The approach we pursue differs from some of the prev..
    corecore